Apple has acquired Xnor.ai, a Seattle startup specializing in low-power, edge-based artificial intelligence tools
https://www.geekwire.com/2020/exclusive-apple-acquires-xnor-ai-edge-ai-spin-paul-allens-ai2-price-200m-range/
More about xnor.ai
https://www.embedded-vision.com/what-embedded-vision/embedded-vision-academy/provider/xnor
https://www.geekwire.com/2020/exclusive-apple-acquires-xnor-ai-edge-ai-spin-paul-allens-ai2-price-200m-range/
More about xnor.ai
https://www.embedded-vision.com/what-embedded-vision/embedded-vision-academy/provider/xnor
GeekWire
Exclusive: Apple acquires Xnor.ai, edge AI spin-out from Paul Allenโs AI2, for price in $200M range
Apple has acquired Xnor.ai, a Seattle startup specializing in low-power, edge-based artificial intelligence tools, sources with knowledge of the deal told GeekWire.
AI chip-design startup EdgeCortix brings total funding to US$4 Million with extended seed round
https://www.linkedin.com/pulse/ai-chip-design-startup-edgecortix-brings-total-funding-dasgupta
https://www.linkedin.com/pulse/ai-chip-design-startup-edgecortix-brings-total-funding-dasgupta
LinkedIn
AI chip-design startup EdgeCortix brings total funding to US$4 Million with extended seed round
(January 2020) The Singapore & Tokyo based startup EdgeCortix Inc. has successfully raised a follow-on seed B of US $1Million in January 2020 from strategic US based investors.
Forwarded from gate-array.com news
Dear Friends, every month we publish a small report that collects all the latest information from the FPGA world. Today we present you the report for January 2020. We collected for you more than 70 news about FPGA!!! Download a PDF of the report http://gate-array.com/reports/Report_2020-01-gate-array.com_digest.pdf
Forwarded from gate-array.com news
Report 2020-01 - gate-array.com digest .pdf
448.8 KB
>>At Intel, we believe that the best approach to creating DSAs ( domain specific architectures) in the data center โ the way that Microsoft has implemented its cloud-based, deep-learning platform for real-time AI inference called Project Brainwave โ is to use overlays. Microsoftโs Project Brainwave employs a soft Neural Processing Unit (NPU) implemented with a high-performance Intelยฎ At Intel, we believe that the best approach to creating DSAs in the data center โ the way that Microsoft has implemented its cloud-based, deep-learning platform for real-time AI inference called Project Brainwave โ is to use overlays. Microsoftโs Project Brainwave employs a soft Neural Processing Unit (NPU) implemented with a high-performance Intelยฎ FPGA to accelerate deep neural network (DNN) inferencing. This NPU DSA has multiple applications in computer vision and natural language processing. to accelerate deep neural network (DNN) inferencing. This NPU DSA has multiple applications in computer vision and natural language processing.
Programmable Logic
The New Golden Age for Computer Architecture Demands FPGA Reconfigurability - Programmable Logic
By Jose Alvarez, Senior Director, Intel CTO Office, PSG The following blog is adapted from a keynote speech that Jose Alvarez presented as the recent The Next FPGA Platform event, held in San Jose, California. Professors John Hennessy and Dave Pattersonโฆ
The toolkit speeds up inference on a broad range of Intel hardware: Intel Xeon Scalable and Core CPUs for general-purpose compute, Intel Movidius VPUs for dedicated media and vision applications, and FPGAs for flexible programming logic and scale.
https://www.intel.ai/open-vino-low-precision-pipeline/#gs.vpoihp
https://www.intel.ai/open-vino-low-precision-pipeline/#gs.vpoihp
Intel
Enhanced low-precision pipeline to accelerate inference
The new release of Intel Distribution of OpenVINO Toolkit features improvements that further optimize deep learning performance.
https://www.coursera.org/learn/comparch/
For all those interested in Computer Architectures - have a look at this course from Princeton.
Some of the covered topics:
Superscalar
Out-of-Order execution
Branch Prediction
Cache coherency
Vector processors
GPUs
Multiprocessor interconnect
For all those interested in Computer Architectures - have a look at this course from Princeton.
Some of the covered topics:
Superscalar
Out-of-Order execution
Branch Prediction
Cache coherency
Vector processors
GPUs
Multiprocessor interconnect
Coursera
Computer Architecture
Offered by Princeton University. In this course, you ... Enroll for free.
The chapter from the upcoming Vivienne Sze book " Efficient Processing of Deep Neural Networks" http://eyeriss.mit.edu/2020_efficient_dnn_excerpt.pdf
* Processing Near Memory
* Processing in memory
* Processing in the Optical Domain
* Processing in Sensor
* Processing Near Memory
* Processing in memory
* Processing in the Optical Domain
* Processing in Sensor
AI on steroids
Regarding hardware, Hinton went into an extended explanation of the technical aspects that constrain today's neural networks. The weights of a neural network, for example, have to be used hundreds of times, he pointed out, making frequent, temporary updates to the weights. He said the fact graphics processing units (GPUs) have limited memory for weights and have to constantly store and retrieve them in external DRAM is a limiting factor.
Much larger on-chip memory capacity "will help with things like Transformer, for soft attention," said Hinton, referring to the wildly popular autoregressive neural network developed at Google in 2017. Transformers, which use "key/value" pairs to store and retrieve from memory, could be much larger with a chip that has substantial embedded memory.
Zdnet
Regarding hardware, Hinton went into an extended explanation of the technical aspects that constrain today's neural networks. The weights of a neural network, for example, have to be used hundreds of times, he pointed out, making frequent, temporary updates to the weights. He said the fact graphics processing units (GPUs) have limited memory for weights and have to constantly store and retrieve them in external DRAM is a limiting factor.
Much larger on-chip memory capacity "will help with things like Transformer, for soft attention," said Hinton, referring to the wildly popular autoregressive neural network developed at Google in 2017. Transformers, which use "key/value" pairs to store and retrieve from memory, could be much larger with a chip that has substantial embedded memory.
Zdnet
NXP Announces Lead Partnership for Arm Ethos-U55 Neural Processing Unit for Machine Learning
https://media.nxp.com/news-releases/news-release-details/nxp-announces-lead-partnership-arm-ethos-u55-neural-processing/
The Ethos-U55 is specifically designed to accelerate ML inference in area-constrained embedded and IoT devices. Its advanced compression techniques save power and reduce ML model sizes significantly to enable execution of neural networks that previously only ran on larger systems. In addition, a unified toolchain with Cortex-M gives developers a simplified, seamless path to develop ML applications within the familiar Cortex-M development environment. The end-to-end enablement, from training to run-time inference deployment for Ethos-U55, will be accessible through NXPโs eIQ machine learning development environment.
https://media.nxp.com/news-releases/news-release-details/nxp-announces-lead-partnership-arm-ethos-u55-neural-processing/
The Ethos-U55 is specifically designed to accelerate ML inference in area-constrained embedded and IoT devices. Its advanced compression techniques save power and reduce ML model sizes significantly to enable execution of neural networks that previously only ran on larger systems. In addition, a unified toolchain with Cortex-M gives developers a simplified, seamless path to develop ML applications within the familiar Cortex-M development environment. The end-to-end enablement, from training to run-time inference deployment for Ethos-U55, will be accessible through NXPโs eIQ machine learning development environment.
NXP Semiconductors - Newsroom
NXP Announces Lead Partnership for Arm Ethos-U55 Neural Processing Unit for Machine Learning | NXP Semiconductors - Newsroom
NXP Semiconductors today announced its lead partnership for the Arm ยฎ Ethos โข -U55 microNPU (Neural Processing Unit), a machine learning (ML) processor targeted at resource-constrained industrial and IoT edge devices. As an industry-leading innovator of โฆ
Graphcore has now raised over $450 million and says that it has some $300 million in cash reserves โ an important detail considering the doldrums that have plagued the chipmaking market in the last few months, and could become exacerbated now with the slowdown in production due to the coronavirus outbreak.
The funding is an extension of its Series D, it said, and brings the total valuation of the company to $1.95 billion.
The funding is an extension of its Series D, it said, and brings the total valuation of the company to $1.95 billion.
The Linley Spring Processor Conference will be held on April 7 - 8, 2020 at the Hyatt Regency Hotel, Santa Clara, California. This two-day conference features in-depth technical presentations addressing processors and IP cores for AI applications, embedded, data center, automotive, and communications. This unique forum focuses on the processors and IP cores used in deep learning, embedded, communications, automotive, IoT, and server designs.
https://www.linleygroup.com/events/event.php?num=48
Participants :
Intel, ARM, SiFive,CEVA, Groq,Cerebras Systems, Mellanox, Cadence
https://www.linleygroup.com/events/event.php?num=48
Participants :
Intel, ARM, SiFive,CEVA, Groq,Cerebras Systems, Mellanox, Cadence
Linleygroup
The Linley Group - Linley Spring Processor Conference 2020
The Linley Spring Processor Conference will be held on April 6-9, 2020. Learn everything you need to know about processors and IP cores used in AI applications, IoT, embedded, data center, automotive, communications, and server designs.
Here we demonstrate that an image sensor can itself constitute an ANN that can simultaneously sense and process optical images without latency. Our device is based on a reconfgurable two-dimensional (2D) semiconductor photodiode array, and the synaptic weights of the network are stored in a continuously tunable photoresponsivity matrix. We demonstrate both supervised and unsupervised learning and train the sensor to classify and encode images that are optically projected onto the chip with a throughput of 20 million bins per second.